Dimension Reduction in Kernel Spaces from Locality-Sensitive Hashing
نویسندگان
چکیده
We provide novel methods for efficient dimensionality reduction in kernel spaces. That is, we provide efficient and explicit randomized maps from “data spaces” into “kernel spaces” of low dimension, which approximately preserve the original kernel values. The constructions are based on observing that such maps can be obtained from Locality-Sensitive Hash (LSH) functions, a primitive developed for fast approximate nearest neighbor search. Thus, we relate the question of dimensionality reduction in kernel spaces to the already existing theory of LSH functions. Efficient dimensionality reduction in kernel spaces enables a substantial speedup of kernelbased algorithms, as experimentally shown in Rahimi-Recht (NIPS’07). Our framework generalizes one of their constructions.
منابع مشابه
Valiant Metric Embeddings , Dimension Reduction
In the previous lecture notes, we saw that any metric (X, d) with |X| = n can be embedded into R 2 n) under any the `1 metric (actually, the same embedding works for any `p metic), with distortion O(log n). Here, we describe an extremely useful approach for reducing the dimensionality of a Euclidean (`2) metric, while incurring very little distortion. Such dimension reduction is useful for a nu...
متن کاملSuper-Bit Locality-Sensitive Hashing
Sign-random-projection locality-sensitive hashing (SRP-LSH) is a probabilistic dimension reduction method which provides an unbiased estimate of angular similarity, yet suffers from the large variance of its estimation. In this work, we propose the Super-Bit locality-sensitive hashing (SBLSH). It is easy to implement, which orthogonalizes the random projection vectors in batches, and it is theo...
متن کاملFast Image Search with Locality-Sensitive Hashing and Homogeneous Kernels Map
Fast image search with efficient additive kernels and kernel locality-sensitive hashing has been proposed. As to hold the kernel functions, recent work has probed methods to create locality-sensitive hashing, which guarantee our approach's linear time; however existing methods still do not solve the problem of locality-sensitive hashing (LSH) algorithm and indirectly sacrifice the loss in accur...
متن کاملIntelligent Control of a Sensor-Actuator System via Kernelized Least-Squares Policy Iteration
In this paper a new framework, called Compressive Kernelized Reinforcement Learning (CKRL), for computing near-optimal policies in sequential decision making with uncertainty is proposed via incorporating the non-adaptive data-independent Random Projections and nonparametric Kernelized Least-squares Policy Iteration (KLSPI). Random Projections are a fast, non-adaptive dimensionality reduction f...
متن کاملDrakkar: a graph based All-Nearest Neighbour Search Algorithm for Bibliographic Coupling
Drakkar is a novel algorithm for the creation of bibliographic coupling graphs in huge document spaces. The algorithm approaches this as an All-Nearest Neighbour search problem and starts from a bipartite graph constituted by the citing publications and the cited references and the directed citations connecting them. The approach is inspired by dimensionality reduction techniques like Random Pr...
متن کامل